Skip to content

Conversation

@ggerganov
Copy link
Member

fix #15215 #15104

When seq_id == -1 we remove the tokens from all KV cache buffers.

@ggerganov ggerganov merged commit 228f724 into master Aug 11, 2025
53 of 55 checks passed
@ggerganov ggerganov deleted the gg/kv-cache-fix-seq-rm branch August 11, 2025 10:58
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

Misc. bug: llama-kv-cache-unified.cpp:222: GGML_ASSERT(seq_id >= 0 && (size_t) seq_id < seq_to_stream.size()) failed when loading processed prompt again

2 participants